21 research outputs found

    A Second-Order Boundary Value Problem with Nonlinear and Mixed Boundary Conditions: Existence, Uniqueness, and Approximation

    Get PDF
    A second-order boundary value problem with nonlinear and mixed two-point boundary conditions is considered, Lx=f(t,x,x′), t∈(a,b), g(x(a),x(b),x′(a),x′(b))=0, x(b)=x(a) in which L is a formally self-adjoint second-order differential operator. Under appropriate assumptions on L, f, and g, existence and uniqueness of solutions is established by the method of upper and lower solutions and Leray-Schauder degree theory. The general quasilinearization method is then applied to this problem. Two monotone sequences converging quadratically to the unique solution are constructed

    Stable Learning via Sample Reweighting

    Full text link
    We consider the problem of learning linear prediction models with model misspecification bias. In such case, the collinearity among input variables may inflate the error of parameter estimation, resulting in instability of prediction results when training and test distributions do not match. In this paper we theoretically analyze this fundamental problem and propose a sample reweighting method that reduces collinearity among input variables. Our method can be seen as a pretreatment of data to improve the condition of design matrix, and it can then be combined with any standard learning method for parameter estimation and variable selection. Empirical studies on both simulation and real datasets demonstrate the effectiveness of our method in terms of more stable performance across different distributed data.Comment: Accepted as poster paper at AAAI202

    Distributionally Robust Learning with Stable Adversarial Training

    Full text link
    Machine learning algorithms with empirical risk minimization are vulnerable under distributional shifts due to the greedy adoption of all the correlations found in training data. There is an emerging literature on tackling this problem by minimizing the worst-case risk over an uncertainty set. However, existing methods mostly construct ambiguity sets by treating all variables equally regardless of the stability of their correlations with the target, resulting in the overwhelmingly-large uncertainty set and low confidence of the learner. In this paper, we propose a novel Stable Adversarial Learning (SAL) algorithm that leverages heterogeneous data sources to construct a more practical uncertainty set and conduct differentiated robustness optimization, where covariates are differentiated according to the stability of their correlations with the target. We theoretically show that our method is tractable for stochastic gradient-based optimization and provide the performance guarantees for our method. Empirical studies on both simulation and real datasets validate the effectiveness of our method in terms of uniformly good performance across unknown distributional shifts.Comment: arXiv admin note: substantial text overlap with arXiv:2006.0441

    Relaxation Oscillations in Singularly Perturbed Generalized Lienard Systems with Non-Generic Turning Points

    Get PDF
    Based on the asymptotic analysis technique developed by Eckhaus [Lecture Notes in Math., vol. 985, pp 449-494. Springer, Berlin, 1983], this paper aims to study the existence and the asymptotic behaviors of relaxation oscillations of regular and canard types in a singularly perturbed generalized Lionard system with a non-generic turning point. The singularly perturbed Lionard system considered in this paper is very general and numerous real world models like some biological ones can be rewritten in the form of this system after a series of transformations. Under certain conditions, we rigorously prove the existence of regular relaxation oscillations and canard relaxation oscillations under the specific parameter conditions. As an application, two biological models, namely, a FitzHugh-Nagumo model and a twodimensional predator-prey model with Holling-II response are studied, in which, the existence of regular relaxation oscillations and canard relaxation oscillations as well as the bifurcation curves are obtained

    Stable Learning via Sample Reweighting

    No full text
    We consider the problem of learning linear prediction models with model misspecification bias. In such case, the collinearity among input variables may inflate the error of parameter estimation, resulting in instability of prediction results when training and test distributions do not match. In this paper we theoretically analyze this fundamental problem and propose a sample reweighting method that reduces collinearity among input variables. Our method can be seen as a pretreatment of data to improve the condition of design matrix, and it can then be combined with any standard learning method for parameter estimation and variable selection. Empirical studies on both simulation and real datasets demonstrate the effectiveness of our method in terms of more stable performance across different distributed data
    corecore